Contrastive Deep Supervision

نویسندگان

چکیده

AbstractThe success of deep learning is usually accompanied by the growth in neural network depth. However, traditional training method only supervises at its last layer and propagates supervision layer-by-layer, which leads to hardship optimizing intermediate layers. Recently, has been proposed add auxiliary classifiers layers networks. By these with supervised task loss, can be applied shallow directly. conflicts well-known observation that learn low-level features instead task-biased high-level semantic features. To address this issue, paper proposes a novel framework named Contrastive Deep Supervision, augmentation-based contrastive learning. Experimental results on nine popular datasets eleven models demonstrate effects general image classification, fine-grained classification object detection learning, semi-supervised knowledge distillation. Codes have released .

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Deep Supervision with Intermediate Concepts

Recent data-driven approaches to scene interpretation predominantly pose inference as an end-to-end black-box mapping, commonly performed by a Convolutional Neural Network (CNN). However, decades of work on perceptual organization in both human and machine vision suggests that there are often intermediate representations that are intrinsic to an inference task, and which provide essential struc...

متن کامل

Training Deeper Convolutional Networks with Deep Supervision

One of the most promising ways of improving the performance of deep convolutional neural networks is by increasing the number of convolutional layers. However, adding layers makes training more difficult and computationally expensive. In order to train deeper networks, we propose to add auxiliary supervision branches after certain intermediate layers during training. We formulate a simple rule ...

متن کامل

Learning Deep Energy Models: Contrastive Divergence vs. Amortized MLE

We propose a number of new algorithms for learning deep energy models from data motivated by a recent Stein variational gradient descent (SVGD) algorithm, including a Stein contrastive divergence (SteinCD) that integrates CD with SVGD based on their theoretical connections, and a SteinGAN that trains an auxiliary generator to generate the negative samples in maximum likelihood estimation (MLE)....

متن کامل

Deep generative-contrastive networks for facial expression recognition

As the expressive depth of an emotional face differs with individuals, expressions, or situations, recognizing an expression using a single facial image at a moment is difficult. One of the approaches to alleviate this difficulty is using a video-based method that utilizes multiple frames to extract temporal information between facial expression images. In this paper, we attempt to utilize a ge...

متن کامل

Towards Lifelong Self-Supervision: A Deep Learning Direction for Robotics

Despite outstanding success in vision amongst other domains, many of the recent deep learning approaches have evident drawbacks for robots. This manuscript surveys recent work in the literature that pertain to applying deep learning systems to the robotics domain, either as means of estimation or as a tool to resolve motor commands directly from raw percepts. These recent advances are only a pi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2022

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-031-19809-0_1